Robust regression via error tolerance

نویسندگان

چکیده

Abstract Real-world datasets are often characterised by outliers; data items that do not follow the same structure as rest of data. These outliers might negatively influence modelling In analysis it is, therefore, important to consider methods robust outliers. this paper we develop a regression method finds largest subset can be approximated using sparse linear model given precision. We show yield best possible robustness However, problem is NP-hard and solve present an efficient approximation algorithm, termed SLISE. Our extends existing state-of-the-art methods, especially in terms speed on high-dimensional datasets. demonstrate our applying both synthetic real-world problems.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Robust Regression via Hard Thresholding

We study the problem of Robust Least Squares Regression (RLSR) where several response variables can be adversarially corrupted. More specifically, for a data matrix X ∈ Rp×n and an underlying model w∗, the response vector is generated as y = XTw∗+b where b ∈ R is the corruption vector supported over at most C ·n coordinates. Existing exact recovery results for RLSR focus solely on L1-penalty ba...

متن کامل

Robust Regression via Hard Thresholding

We study the problem of Robust Least Squares Regression (RLSR) where several response variables can be adversarially corrupted. More specifically, for a data matrix X ∈ Rp×n and an underlying model w∗, the response vector is generated as y = XTw∗ +b where b ∈ R is the corruption vector supported over at most C · n coordinates. Existing exact recovery results for RLSR focus solely on L1penalty b...

متن کامل

Relative error prediction via kernel regression smoothers

In this article, we introduce and study local constant and our preferred local linear nonparametric regression estimators when it is appropriate to assess performance in terms of mean squared relative error of prediction. We give asymptotic results for both boundary and non-boundary cases. These are special cases of more general asymptotic results that we provide concerning the estimation of th...

متن کامل

Robust Subspace Clustering via Thresholding Ridge Regression

In this material, we provide the theoretical analyses to show that the trivial coefficients always correspond to the codes over errors. Lemmas 1–3 show that our errors-removing strategy will perform well when the lp-norm is enforced over the representation, where p = {1, 2,∞}. Let x 6= 0 be a data point in the union of subspaces SD that is spanned by D = [Dx D−x], where Dx and D−x consist of th...

متن کامل

Robust Regression via Heuristic Hard Thresholding

The presence of data noise and corruptions recently invokes increasing attention on Robust Least Squares Regression (RLSR), which addresses the fundamental problem that learns reliable regression coefficients when response variables can be arbitrarily corrupted. Until now, several important challenges still cannot be handled concurrently: 1) exact recovery guarantee of regression coefficients 2...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Data Mining and Knowledge Discovery

سال: 2022

ISSN: ['1573-756X', '1384-5810']

DOI: https://doi.org/10.1007/s10618-022-00819-2